A conditional gradient method with linear rate of convergence for solving convex linear systems

نویسندگان

  • Amir Beck
  • Marc Teboulle
چکیده

We consider the problem of finding a point in the intersection of an affine set with a compact convex set, called a convex linear system (CLS). The conditional gradient method is known to exhibit a sublinear rate of convergence. Exploiting the special structure of (CLS), we prove that the conditional gradient method applied to the equivalent minimization formulation of (CLS), converges to a solution at a linear rate, under the sole assumption that Slater’s condition holds for (CLS). The rate of convergence is measured explicitly in terms of the problem’s data and a Slater point. Application to a class of conic linear systems is discussed.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Comparison results on the preconditioned mixed-type splitting iterative method for M-matrix linear systems

Consider the linear system Ax=b where the coefficient matrix A is an M-matrix. In the present work, it is proved that the rate of convergence of the Gauss-Seidel method is faster than the mixed-type splitting and AOR (SOR) iterative methods for solving M-matrix linear systems. Furthermore, we improve the rate of convergence of the mixed-type splitting iterative method by applying a preconditio...

متن کامل

Modify the linear search formula in the BFGS method to achieve global convergence.

<span style="color: #333333; font-family: Calibri, sans-serif; font-size: 13.3333px; font-style: normal; font-variant-ligatures: normal; font-variant-caps: normal; font-weight: 400; letter-spacing: normal; orphans: 2; text-align: justify; text-indent: 0px; text-transform: none; white-space: normal; widows: 2; word-spacing: 0px; -webkit-text-stroke-width: 0px; background-color: #ffffff; text-dec...

متن کامل

On the convergence speed of artificial neural networks in‎ ‎the solving of linear ‎systems

‎Artificial neural networks have the advantages such as learning, ‎adaptation‎, ‎fault-tolerance‎, ‎parallelism and generalization‎. ‎This ‎paper is a scrutiny on the application of diverse learning methods‎ ‎in speed of convergence in neural networks‎. ‎For this aim‎, ‎first we ‎introduce a perceptron method based on artificial neural networks‎ ‎which has been applied for solving a non-singula...

متن کامل

Preconditioned Generalized Minimal Residual Method for Solving Fractional Advection-Diffusion Equation

Introduction Fractional differential equations (FDEs)  have  attracted much attention and have been widely used in the fields of finance, physics, image processing, and biology, etc. It is not always possible to find an analytical solution for such equations. The approximate solution or numerical scheme  may be a good approach, particularly, the schemes in numerical linear algebra for solving ...

متن کامل

An accelerated gradient based iterative algorithm for solving systems of coupled generalized Sylvester-transpose matrix equations

‎In this paper‎, ‎an accelerated gradient based iterative algorithm for solving systems of coupled generalized Sylvester-transpose matrix equations is proposed‎. ‎The convergence analysis of the algorithm is investigated‎. ‎We show that the proposed algorithm converges to the exact solution for any initial value under certain assumptions‎. ‎Finally‎, ‎some numerical examples are given to demons...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Math. Meth. of OR

دوره 59  شماره 

صفحات  -

تاریخ انتشار 2004